122 research outputs found

    A Future Data Environment - reusability vs. citability and synchronisation vs. ingestion

    Get PDF
    During the last decades data managers dedicated their work to the pursuit for importable data. In the recent years this chase seams to come to an end while funding organisations assume that the approach of data publications with citable data sets will eliminate denial of scientists to commit their data. But is this true for all problems we are facing at the edge of a data avalanche and data intensive science? The concept of citable data is a logical consequence from the connection of points. Potential data providers in the past complained usually about the missing of a credit assignment for data providers and they still do. The selected way of DOI captured data sets is perfectly fitting into the credit system of publisher driven publications with countable citations. This system is well known by scientists for approximately 400 years now. Unfortunately, there is a double bind situation between citeability and reusability. While cooperation of publishers and data archives are coming into existence, it is necessary to get one question clear: “Is it really worth while in the twenty-first century to force data into the publication process of the seventeenth century?” Data publications enable easy citability, but do not support easy data reusability for future users. Additional problems occur in such an environment while taking into account the chances of collaborative data corrections in the institutional repository. The future with huge amounts of data connected with publications makes reconsideration towards a more integrated approach reasonable. In the past data archives were the only infrastructures taking care of long-term data retrievability and availability. Nevertheless, they were never a part of the scientific process from data creation, analysis, interpretation and publication. Data archives were regarded as isolated islands in the sea of scientific data. Accordingly scientists considered data publications like a stumbling stone in their daily routines and still do. The creation of data set as additional publications is an additional workload a lot of scientists are not yet convinced about. These times are coming to an end now because of the efforts of the funding organisations and the increased awareness of scientific institutions. Right now data archives have their expertise in retrievability and availability, but the new demand of data provenance is not yet included in their systems. So why not taking the chance of the scientific institutes sneaking in and split the workload of retrievability and provenance. Such an integrated data environment will be characterized by increased functionality, creditability and structured data from the creation and everything accompanied by data managers. The Kiel Data Management Infrastructure is creating such an institutional provenance system for the scientific site of Kiel. Having data sets up to date by synchronisation with institutional provenance system capturing all changes and improvements right where they happen. A sophisticated and scalable landscape needs to combine advantages of the existing data centers such as the usability and retrievability functionality with the advantages of decentralised data capturing and provenance. This data environment with synchronisation features and creditability of scientific data to future users would be capable of the future tasks

    Kiel Data Management Infrastructure for Marine Sciences

    Get PDF

    Author identities an interoperability problem solved by a collaborative solution

    Get PDF
    The identity of authors and data providers is crucial for personalized interoperability. The marketplace of available identifiers is packed and the right choice is getting more and more complicated. Even though there are more then 15 different systems available there are still some under development and proposed to come up by the end of 2012 ('PubMed Central Author ID' and ORCID). Data Management on a scale beyond the size of a single research institute but on the scale of a scientific site including a university with student education program needs to tackle this problem and so did the Kiel Data Management an Infrastructure. The main problem with the identities of researchers is the quite high frequency changes in positions during a scientist life. The required system needed to be a system that already contained the potential of preregistered people with their scientific publications from other countries, institutions and organizations. Scanning the author ID marketplace brought up, that there us a high risk of additional workload to the researcher itself or the administration due to the fact that individuals need to register an ID for themselves or the chosen register is not yet big enough to simply find the right entry. On the other hand libraries deal with authors and their publications now for centuries and they have high quality catalogs with person identities already available. Millions of records internationally mapped are available by collaboration with libraries and can be used in exactly the same scope. The international collaboration between libraries (VIAF) provides a mapping between libraries from the US, CA, UK, FR, GER and many more. The international library author identification system made it possible to actually reach at the first matching a success of 60% of all scientists. The additional advantage is that librarians can finalize the Identity system in a kind of background process. The Kiel Data Management Infrastructure initiated a web service at Kiel for mapping from one ID to another. This web service supports the scientific workflows for automation of the data archiving process at world data archive PANGAEA. The long-lasting concept of the library identifier enables the use of these identifiers beyond the employment period, while it has nothing to do with the institutional IDM. The access rights and ownership of data can be assured for very long time since the national library with its national scope hosts the basic system. Making use of this existing system released resourced planed for this task and enabled the chance of interoperability on an international scale for a regional data management infrastructure

    HowTo - Easy use of global unique identifier

    Get PDF
    Czerniak A, Schirnick C, Fleischer D, Mehrtens H, Springer P. HowTo - Easy use of global unique identifier. Presented at the AGU Fall Meeting 2013, San Francisco, US.First Step We use our scientifi c workfl ow system to defi ne the research workfl ow step by step. In every workstep metadata information is used to prebind to the proper identifi er system (eg. ports from geographic places and not people) which can then propose adequate suggestions upon user input leading to the correct assignment of the port in question. Sometimes a mapping between identifi er resources such as the german " Gemeinsame Normdatei " (GND) for people , institutions or geographic places and another like ISNI or OrcID for people only may be necessary for legacy reasons. For example our handwriting recognition system in conjunction with " digital enabled " paper uses names and maps to relate those to the respective systems in order to retrieve a global unique identifi er which is then used to reliably relate and store information (Figure 2). Identifi er system are ubiquitous but commonly they are not global(ly known) nor are they unique (national vs. international). Usage of global identifi ers in marine research greatly enhances search and retrieval of samples, data and related information and enbales scientists all over the world to reliably share and reuse research data. For many scientifi c disciplines numerous identifi er systems exists on a national and international level allowing non-ambiguous refernce to people, institutions and geosamples (Figure 1). Our goal is to reliably integrate these unique references in our virtual research environment, the Kiel Data Management Infrastructure (KDMI), supporting scientists in national and international projects and collaborative research centers

    Data Management: The Data Life Cycle

    Get PDF
    The scientific site of Kiel provides support for projects with data management requirements due to project size or interdisciplinarity. This infrastructure is the Kiel Data Management Infrastructure (KDMI) and was initially created by SFB574, SFB754, Excellence Cluster ‘The Future Ocean‘ and the GEOMAR | Helmholtz Centre for Ocean Research Kiel. To achieve public data availability from publicly funded projects by the end of the funding period it is necessary to initiate the data acquisition during the data creation process. Accordingly the KDMI uses a three level approach to achieve this goal in SOPRAN III. Data management is al- ready involvedin the planning of expeditions or experiments. The resulting schedule for data files can be used by the project coordinationto increase the efficeny of data sharing within SOPRAN III. The scientists provide files with basic metainformation, which are available within the virtual research environment as soon as possible to all project members. Final data will be transferred to PANGAEA for long term availability when the data are analysed and interpreted in a scientific publication or by the end of SOPRAN III. The Kiel Data Management Team offers a portal for all GEOMAR and University Kiel marine projects. This portal will be used in SOPRAN III in combination with PANGAEA to fulfill the project’s data management requirements and to enhance the data sharing within SOPRAN III by a file sharing environment for preliminary data not yet suitable for PANGAEA

    The next generation of data capturing - digital ink for the data stewards of the future

    Get PDF
    Czerniak A, Fleischer D, Schirnick C, Springer P, Mehrtens H. The next generation of data capturing - digital ink for the data stewards of the future. Presented at the AGU Fall Meeting 2012, San Francisco.The Kiel Data Management (KDM) Infrastructure places data capturing straight in the data creation process. With this approach student education courses can be just another use case of data capturing. Smoothing this process and usage of available technologies inspired the KDM Team to prototype with " Smart Pen " and digital Ink. This technology with handwriting recognition fi lls the gap between the very long-lasting paper protocols and their manual digitalization of fi eld and sampling data. The combination of the KDM Infrastructure with the digital Ink technology enables data capturing from student education throughout high-end scientifi c lab work. The realization of persistent information started approx. 37000 years ago in the days of cave paintings (Fig. 1). In the following history stone tablets (eg. sPad) were used to record information fi rst and then a steady development of persistent information from " paper and ink feather " , " paper and printing techniques " to " computer-based information recording " happened. Until " computer-based information recording " , the paper-based recording was the choice throughout the last centuries. In many scientifi c disciplines paper-based recording remained the only option over the last decades. These data and informations need to be converted to a digital representation for further " computer-based " use. After these manual and error-prone transformations the information can be analyzed and interpreted in the context of scientifi c research

    Rock and Core Repository Coming Digital

    Get PDF
    In times of whole city centres being available by a mouse click in 3D to virtually walk through, reality sometimes becomes neglected. The reality of scientific sample collections not being digitised to the essence of molecules, isotopes and electrons becomes unbelievable to the upgrowing generation of scientists. Just like any other geological institute the Helmholtz Centre for Ocean Research GEOMAR accumulated thousands of specimen. The samples, collected mainly during marine expeditions, date back as far as 1964. Today GEOMAR houses a central geological sample collection of at least 17 000 m of sediment core and more than 4 500 boxes with hard rock samples and refined sample specimen. This repository, having been dormant, missed the onset of the interconnected digital age. Physical samples without barcodes, QR codes or RFID tags need to be migrated and reconnected, urgently. In our use case, GEOMAR opted for the International Geo Sample Number IGSN as the persistent identifier. Consequentially, the software CurationDIS by smartcube GmbH as the central component of this project was selected. The software is designed to handle acquisition and administration of sample material and sample archiving in storage places. In addition, the software allows direct embedding of IGSN. We plan to adopt IGSN as a future asset, while for the initial inventory taking of our sample material, simple but unique QR codes act as “bridging identifiers” during the process. Currently we compile an overview of the broad variety of sample types and their associated data. QR-coding of the boxes of rock samples and sediment cores is near completion, delineating their location in the repository and linking a particular sample to any information available about the object. Planning is in progress to streamline the flow from receiving new samples to their curation to sharing samples and information publically. Additionally, interface planning for linkage to GEOMAR databases OceanRep (publications) and OSIS (expeditions) as well as for external data retrieval are in the pipeline. Looking ahead to implement IGSN, taking on board lessons learned from earlier generations, it will enable to comply with our institute’s open science policy. Also it will allow to register newly collected samples already during ship expeditions. They thus receive their "birth certificate" contemporarily in this ever faster revolving scientific world

    FAIR-Prinzipien und Tools fĂŒr die Evaluation und das Assessment

    Get PDF
    FAIR steht im wissenschaftlichen Kontext fĂŒr das Akronym Findable, Accessible, Interoperable und Reusable. Der Vortrag gibt im ersten Teil einen kurzen theoretischen Überblick ĂŒber die FAIR-Prinzipien und stellt die AktivitĂ€ten der Research Data Alliance (RDA) vor. FAIR-Projekte wie FAIRsFAIR, GOFAIR, aber auch die AktivitĂ€ten der European Open Science Cloud (EOSC) werden behandelt. Im zweiten Teil werden zwei existierende Tools zur UnterstĂŒtzung der Evaluierung exemplarisch vorgestellt. F-UJI aus dem FAIRsFAIR Projekt und der OpenAIRE FAIR-Assistant Validator.Der Vortrag scließt mit einem Ausblick auf aktuelle Entwicklungen in der FAIR-Community und den Evaluator-Tools

    OpenAIRE-Nexus Scholarly Communication Services for EOSC users - ein abschließender Blick

    Get PDF
    Das EU-finanzierte Projekt "OpenAIRE-Nexus - Eine zugĂ€nglichere offene Wissenschaft" (https://cordis.europa.eu/project/id/101017452/de) hatte zum Ziel, in der Welt der offenen Wissenschaft eine faire, benutzerfreundliche und offene Kommunikation sowie Informationsweitergabe zu fördern. Um Forschung zu unterstĂŒtzen, mĂŒssen Inhalte und Informationen daher leicht zu finden und anzuzeigen sein. OpenAIRE Nexus stellte ein Diensteportfolio zusammen, das bei der Veröffentlichung von Forschungsergebnissen, zum Monitoring ihres Impacts und zur Verbreitung des neuen Wissens wesentlich beitragen kann. Diese in Europa und darĂŒber hinaus bereits etablierten Dienstleistungen werden von öffentlich-rechtlichen Institutionen und Einrichtungen bereitgestellt. Ziel war es, dieses Portfolio in die European Open Science Cloud (EOSC) und die globale offene Wissenschaftsgemeinschaft einzubinden. Der erste Teil des Vortrages wird eine Übersicht ĂŒber das Projekt und dessen Ziele geben. Im zweiten Teil werden beispielhaft Ergebnisse des Projektes, insbesondere solcher Arbeitspakete, an denen die UniversitĂ€tsbibliothek Bielefeld beteiligt war, prĂ€sentiert. Dabei handelt sich unter anderem um die Integration von OpenAPC in den OpenAIRE Research Graphen, sowie die Bekanntmachung von OpenAPC im internationalen Kontext. Arbeiten an der Integration von Forschungsinformationssystemen (FISs), sowie Open Research Europe (ORE, https://open-research-europe.ec.europa.eu/) - der Open Access Publishing Plattform der EuropĂ€ischen Kommission, gehörten gleichfalls zu den erfolgreich abgeschlossen Aufgaben

    FAIR-Prinzipien und Tools fĂŒr die Evaluation und das Assessment

    Get PDF
    FAIR steht im wissenschaftlichen Kontext fĂŒr das Akronym Findable, Accessible, Interoperable und Reusable. Der Vortrag gibt im ersten Teil einen kurzen theoretischen Überblick ĂŒber die FAIR-Prinzipien und stellt die AktivitĂ€ten der Research Data Alliance (RDA) vor. FAIR-Projekte wie FAIRsFAIR, GOFAIR, aber auch die AktivitĂ€ten der European Open Science Cloud (EOSC) werden behandelt. Im zweiten Teil werden zwei existierende Tools zur UnterstĂŒtzung der Evaluierung exemplarisch vorgestellt. F-UJI aus dem FAIRsFAIR Projekt und der OpenAIRE FAIR-Assistant Validator.Der Vortrag scließt mit einem Ausblick auf aktuelle Entwicklungen in der FAIR-Community und den Evaluator-Tools
    • 

    corecore